Regularization with Approximated L2 Maximum Entropy Method
نویسنده
چکیده
We tackle the inverse problem of reconstructing an unknown finite measure μ from a noisy observation of a generalized moment of μ defined as the integral of a continuous and bounded operator Φ with respect to μ . When only a quadratic approximation Φm of the operator is known, we introduce the L2 approximate maximum entropy solution as a minimizer of a convex functional subject to a sequence of convex constraints. Under several assumptions on the convex functional, the convergence of the approximate solution is established and rates of convergence are provided.
منابع مشابه
Regularization with Approximated L Maximum Entropy Method
We tackle the inverse problem of reconstructing an unknown finite measure μ from a noisy observation of a generalized moment of μ defined as the integral of a continuous and bounded operator Φ with respect to μ . When only a quadratic approximation Φm of the operator is known, we introduce the L2 approximate maximum entropy solution as a minimizer of a convex functional subject to a sequence of...
متن کاملOn maximum entropy regularization for a specific inverse problem of option pricing
We investigate the applicability of the method of maximum entropy regularization (MER) to a specific nonlinear ill-posed inverse problem (SIP) in a purely time-dependent model of option pricing, introduced and analyzed for an L2-setting in [9]. In order to include the identification of volatility functions with a weak pole, we extend the results of [12] and [13], concerning convergence and conv...
متن کاملA Comparative Study of Parameter Estimation Methods for Statistical Natural Language Processing
This paper presents a comparative study of five parameter estimation algorithms on four NLP tasks. Three of the five algorithms are well-known in the computational linguistics community: Maximum Entropy (ME) estimation with L2 regularization, the Averaged Perceptron (AP), and Boosting. We also investigate ME estimation with L1 regularization using a novel optimization algorithm, and BLasso, whi...
متن کاملA non-extensive maximum entropy based regularization method for bad conditioned inverse problems
A regularization method based on the non-extensive maximum entropy principle is devised. Special emphasis is given to the q = 1=2 case. We show that, when the residual principle is considered as constraint, the q = 1=2 generalized distribution of Tsallis yields a regularized solution for bad-conditioned problems. The so devised regularized distribution is endowed with a component which correspo...
متن کاملPanel discussion section I
Kochukhov: The Doppler imaging problem is intrinsically ill-conditioned. It is not correct to characterize this problem by the number of unknowns. We are using regularization, and regularization effectively introduces a set of additional equations which smoothes the map in our case; in the case of the maximum entropy method, it selects an image with maximum entropy. So actually if I count equat...
متن کامل